# Sigmoid loss function

Vit B 16 SigLIP2 384
Apache-2.0
SigLIP 2 vision-language model trained on the WebLI dataset, suitable for zero-shot image classification tasks
Text-to-Image
V
timm
1,497
0
Siglip So400m Patch14 224
Apache-2.0
SigLIP is an improved multimodal model based on CLIP, employing a superior Sigmoid loss function, pre-trained on the WebLI dataset, and suitable for tasks such as zero-shot image classification and image-text retrieval.
Text-to-Image Transformers
S
google
6,654
53
Siglip So400m Patch14 384
Apache-2.0
SigLIP is a vision-language model pre-trained on the WebLi dataset, employing an improved sigmoid loss function to optimize image-text matching tasks.
Image-to-Text Transformers
S
google
6.1M
526
Siglip Base Patch16 256 Multilingual
Apache-2.0
SigLIP is an improved CLIP model pre-trained on the WebLi dataset, optimized for image-text matching tasks using a Sigmoid loss function
Text-to-Image Transformers
S
google
175.86k
44
Siglip Large Patch16 384
Apache-2.0
SigLIP is a multimodal model pretrained on the WebLi dataset, utilizing an improved Sigmoid loss function, suitable for zero-shot image classification and image-text retrieval tasks.
Image-to-Text Transformers
S
google
245.21k
6
Siglip Large Patch16 256
Apache-2.0
SigLIP is a vision-language model pre-trained on the WebLi dataset, utilizing an improved sigmoid loss function to enhance performance
Image-to-Text Transformers
S
google
24.13k
12
Siglip Base Patch16 512
Apache-2.0
SigLIP is a vision-language model pretrained on the WebLi dataset, utilizing an improved sigmoid loss function, excelling in image classification and image-text retrieval tasks.
Text-to-Image Transformers
S
google
237.79k
24
Siglip Base Patch16 384
Apache-2.0
SigLIP is a multimodal model pre-trained on the WebLi dataset, employing an improved sigmoid loss function, suitable for zero-shot image classification and image-text retrieval tasks.
Image-to-Text Transformers
S
google
2,570
10
Siglip Base Patch16 256
Apache-2.0
SigLIP is a vision-language model pre-trained on the WebLi dataset, employing an improved Sigmoid loss function, excelling in image classification and image-text retrieval tasks.
Text-to-Image Transformers
S
google
12.71k
5
Siglip Base Patch16 224
Apache-2.0
SigLIP is a vision-language model pretrained on the WebLi dataset, utilizing an improved Sigmoid loss function to optimize image-text matching tasks
Image-to-Text Transformers
S
google
250.28k
43
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase